Regularized singular value decomposition: a sparse dimension reduction technique

نویسنده

  • Baolin Wu
چکیده

Singular value decomposition (SVD) is a useful multivariate technique for dimension reduction. It has been successfully applied to analyze microarray data, where the eigen vectors are called eigen-genes/arrays. One weakness associated with the SVD is the interpretation. The eigen-genes are essentially linear combinations of all the genes. It is desirable to have sparse SVD, which retains the dimension reduction property but also the eigen vectors are linear combinations of only a small subset of genes. In this paper we formally propose a statistical framework for sparse SVD, which is a generalization of traditional SVD. Our formulation of sparse SVD also bears the close connection to the penalized t-/F-statistics for differential gene expression detection, which can be viewed as the supervised sparse dimension reduction technique.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Sparse Principal Component Analysis via Regularized Low Rank Matrix Approximation

Principal component analysis (PCA) is a widely used tool for data analysis and dimension reduction in applications throughout science and engineering. However, the principal components (PCs) can sometimes be difficult to interpret, because they are linear combinations of all the original variables. To facilitate interpretation, sparse PCA produces modified PCs with sparse loadings, i.e. loading...

متن کامل

Analysis of Singular Value Decomposition as a Dimensionality Reduction Technique and Sparse Representation Classifier as a Post Classifier for the Classification of Epilepsy Risk Levels from EEG Signals

The main aim of this paper is to perform the analysis of Singular Value Decomposition (SVD) as a Dimensionality Reduction technique and Sparse Representation Classifier (SRC) as a Post Classifier for the Classification of Epilepsy Risk levels from Electroencephalography signals. The data acquisition of the EEG signals is performed initially. Then SVD is applied here as a dimensionality reductio...

متن کامل

Large-scale Inversion of Magnetic Data Using Golub-Kahan Bidiagonalization with Truncated Generalized Cross Validation for Regularization Parameter Estimation

In this paper a fast method for large-scale sparse inversion of magnetic data is considered. The L1-norm stabilizer is used to generate models with sharp and distinct interfaces. To deal with the non-linearity introduced by the L1-norm, a model-space iteratively reweighted least squares algorithm is used. The original model matrix is factorized using the Golub-Kahan bidiagonalization that proje...

متن کامل

Fast Linear Discriminant Analysis using QR Decomposition and Regularization

Linear Discriminant Analysis (LDA) is among the most optimal dimension reduction methods for classification, which provides a high degree of class separability for numerous applications from science and engineering. However, problems arise with this classical method when one or both of the scatter matrices is singular. Singular scatter matrices are not unusual in many applications, especially f...

متن کامل

Sparse Principal Component Analysis Incorporating Stability Selection

Principal component analysis (PCA) is a popular dimension reduction method that approximates a numerical data matrix by seeking principal components (PC), i.e. linear combinations of variables that captures maximal variance. Since each PC is a linear combination of all variables of a data set, interpretation of the PCs can be difficult, especially in high-dimensional data. In order to find ’spa...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005